TIL: For long-lived LLM sessions, swapping KV Cache to RAM is ~10x faster than recalculating it. Why isn't this a standard feature?
🦙Simple finetuning LLMs
Flag this post
Cycle-accurate 6502 emulator as coroutine in Rust
🔥Svelte
Flag this post
Opportunistically Parallel Lambda Calculus
🔵LLM frameworks and AI libraries for TypeScript
Flag this post
How fast can an LLM go?
📊Vector Databases
Flag this post
No Cap, This Memory Slaps: Breaking Through the OLTP Memory Wall
🦙Simple finetuning LLMs
Flag this post
DGX Spark UMA can trick you
🔥Svelte
Flag this post
From Lossy to Lossless Reasoning
🤖Coding Automation
Flag this post
Linux/WASM
🧩WASI
Flag this post
Can you save on LLM tokens using images instead of text?
🔵LLM frameworks and AI libraries for TypeScript
Flag this post
I built a lightweight HTTP bridge for AnythingLLM to safely run multiple local MCPs inside Docker (Dummy + Time demo included)
🧩WASI
Flag this post
Loading...Loading more...